Goto

Collaborating Authors

 near-minimax-optimal distributional reinforcement learning


Near-Minimax-Optimal Distributional Reinforcement Learning with a Generative Model

Neural Information Processing Systems

We propose a new algorithm for model-based distributional reinforcement learning (RL), and prove that it is minimax-optimal for approximating return distributions in the generative model regime (up to logarithmic factors), the first result of this kind for any distributional RL algorithm. Our analysis also provides new theoretical perspectives on categorical approaches to distributional RL, as well as introducing a new distributional Bellman equation, the stochastic categorical CDF Bellman equation, which we expect to be of independent interest. Finally, we provide an experimental study comparing a variety of model-based distributional RL algorithms, with several key takeaways for practitioners.


Near-Minimax-Optimal Distributional Reinforcement Learning with a Generative Model

Rowland, Mark, Wenliang, Li Kevin, Munos, Rémi, Lyle, Clare, Tang, Yunhao, Dabney, Will

arXiv.org Artificial Intelligence

We propose a new algorithm for model-based distributional reinforcement learning (RL), and prove that it is minimax-optimal for approximating return distributions with a generative model (up to logarithmic factors), resolving an open question of Zhang et al. (2023). Our analysis provides new theoretical results on categorical approaches to distributional RL, and also introduces a new distributional Bellman equation, the stochastic categorical CDF Bellman equation, which we expect to be of independent interest. We also provide an experimental study comparing several model-based distributional RL algorithms, with several takeaways for practitioners.